Divergence Function, Duality, and Convex Analysis

نویسنده

  • Jun Zhang
چکیده

From a smooth, strictly convex function phi: Rn --> R, a parametric family of divergence function Dphi(alpha) may be introduced: [ equation: see text] for x, y epsilon int dom (Phi) subset Rn, and for alpha in R, with Dphi(+/-1) defined through taking the limit of alpha. Each member is shown to induce an alpha-independent Riemannian metric, as well as a pair of dual alpha-connections, which are generally nonflat, except for alpha = +/-1. In the latter case, Dphi(+/-1) reduces to the (nonparametric) Bregman divergence, which is representable using phi and its convex conjugate phi* and becomes the canonical divergence for dually flat spaces (Amari, 1982, 1985; Amari & Nagaoka, 2000). This formulation based on convex analysis naturally extends the informationgeometric interpretation of divergence functions (Eguchi, 1983) to allow the distinction between two different kinds of duality: referential duality (alpha <--> -alpha) and representational duality (phi <--> phi*). When applied to (not necessarily normalized) probability densities, the concept of conjugated representations of densities is introduced, so that +/-alpha-connections defined on probability densities embody both referential and representational duality and are hence themselves bidual. When restricted to a finite-dimensional affine submanifold, the natural parameters of a certain representation of densities and the expectation parameters under its conjugate representation form biorthogonal coordinates. The alpha representation (indexed by beta now, beta epsilon [-1, 1]) is shown to be the only measure-invariant representation. The resulting two-parameter family of divergence functionals D(alpha,beta), (alpha, beta) epsilon [-1, 1] x [-1, 1] induces identical Fisher information but bidual alpha-connection pairs; it reduces in form to Amari's alpha-divergence family when alpha = +/-1 or when beta = 1, but to the family of Jensen difference (Rao, 1987) when beta = -1.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Nonparametric Information Geometry: From Divergence Function to Referential-Representational Biduality on Statistical Manifolds

Divergence functions are the non-symmetric “distance” on the manifold,Mθ, of parametric probability density functions over a measure space, (X,μ). Classical information geometry prescribes, on Mθ: (i) a Riemannian metric given by the Fisher information; (ii) a pair of dual connections (giving rise to the family of α-connections) that preserve the metric under parallel transport by their joint a...

متن کامل

Optimality and Duality for an Efficient Solution of Multiobjective Nonlinear Fractional Programming Problem Involving Semilocally Convex Functions

In this paper, the problem under consideration is multiobjective non-linear fractional programming problem involving semilocally convex and related functions. We have discussed the interrelation between the solution sets involving properly efficient solutions of multiobjective fractional programming and corresponding scalar fractional programming problem. Necessary and sufficient optimality...

متن کامل

A short variational proof of equivalence between policy gradients and soft Q learning

Two main families of reinforcement learning algorithms, Q-learning and policy gradients, have recently been proven to be equivalent when using a softmax relaxation on one part, and an entropic regularization on the other. We relate this result to the well-known convex duality of Shannon entropy and the softmax function. Such a result is also known as the Donsker-Varadhan formula. This provides ...

متن کامل

t-divergence Based Approximate Inference

Approximate inference is an important technique for dealing with large, intractable graphical models based on the exponential family of distributions. We extend the idea of approximate inference to the t-exponential family by defining a new t-divergence. This divergence measure is obtained via convex duality between the log-partition function of the t-exponential family and a new t-entropy. We ...

متن کامل

Strong Duality in Robust Convex Programming: Complete Characterizations

Abstract. Duality theory has played a key role in convex programming in the absence of data uncertainty. In this paper, we present a duality theory for convex programming problems in the face of data uncertainty via robust optimization. We characterize strong duality between the robust counterpart of an uncertain convex program and the optimistic counterpart of its uncertain Lagrangian dual. We...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Neural computation

دوره 16 1  شماره 

صفحات  -

تاریخ انتشار 2004